Non-Asymptotic Converse Bounds and Refined Asymptotics for Two Lossy Source Coding Problems
نویسندگان
چکیده
In this paper, we revisit two multi-terminal lossy source coding problems: the lossy source coding problem with side information available at the encoder and one of the two decoders, which we term as the Kaspi problem (Kaspi, 1994), and the multiple description coding problem with one semi-deterministic distortion measure, which we refer to as the Fu-Yeung problem (Fu and Yeung, 2002). For the Kaspi problem, we first present the properties of optimal test channels. Subsequently, we generalize the notion of the distortion-tilted information density for the lossy source coding problem to the Kaspi problem and prove a nonasymptotic converse bound using the properties of optimal test channels and the well-defined distortion-tilted information density. Finally, for discrete memoryless sources, we derive refined asymptotics which includes the second-order, large and moderate deviations asymptotics. In the converse proof of second-order asymptotics, we apply the Berry-Esseen theorem to the derived non-asymptotic converse bound. The achievability proof follows by first proving a type-covering lemma tailored to the Kaspi problem, then properly Taylor expanding the well-defined distortion-tilted information densities and finally applying the BerryEsseen theorem. We then generalize the methods used in the Kaspi problem to the Fu-Yeung problem. As a result, we obtain the properties of optimal test channels for the minimum sum-rate function, a non-asymptotic converse bound and refined asymptotics for discrete memoryless sources. Since the successive refinement problem is a special case of the Fu-Yeung problem, as a byproduct, we obtain a non-asymptotic converse bound for the successive refinement problem, which is a strict generalization of the non-asymptotic converse bound for successively refinable sources (Zhou, Tan and Motani, 2017). Index Terms Lossy source coding, multiple description coding problem, Non-asymptotic converse bound, Second-order asymptotics, Large Deviations, Moderate Deviations
منابع مشابه
Cumulant Generating Function of Codeword Lengths in Variable-Length Lossy Compression Allowing Positive Excess Distortion Probability
This paper considers the problem of variable-length lossy source coding. The performance criteria are the excess distortion probability and the cumulant generating function of codeword lengths. We derive a non-asymptotic fundamental limit of the cumulant generating function of codeword lengths allowing positive excess distortion probability. It is shown that the achievability and converse bound...
متن کاملLossy data compression: nonasymptotic fundamental limits
The basic tradeoff in lossy compression is that between the compression ratio (rate) and the fidelity of reproduction of the object that is compressed. Traditional (asymptotic) information theory seeks to describe the optimum tradeoff between rate and fidelity achievable in the limit of infinite length of the source block to be compressed. A perennial question in information theory is how relev...
متن کاملSimple one-shot bounds for various source coding problems using smooth Rényi quantities
We consider the problem of source compression under three different scenarios in the one-shot (nonasymptotic) regime. To be specific, we prove one-shot achievability and converse bounds on the coding rates for distributed source coding, source coding with coded side information available at the decoder and source coding under maximum distortion criterion. The one-shot bounds obtained are in ter...
متن کاملRelative entropies and their use in quantum information theory
This dissertation investigates relative entropies, also called generalized divergences, and how they can be used to characterize information-theoretic tasks in quantum information theory. The main goal is to further refine characterizations of the optimal rates for quantum source coding, state redistribution, and measurement compression with quantum side information via second order asymptotic ...
متن کاملSecond-Order Coding Rates for Conditional Rate-Distortion
This paper characterizes the second-order coding rates for lossy source coding with side information available at both the encoder and the decoder. We first provide non-asymptotic bounds for this problem and then specialize the non-asymptotic bounds for three different scenarios: discrete memoryless sources, Gaussian sources, and Markov sources. We obtain the second-order coding rates for these...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1708.05496 شماره
صفحات -
تاریخ انتشار 2017